A Second Derivative Sqp Method with Imposed
نویسندگان
چکیده
Sequential quadratic programming (SQP) methods form a class of highly efficient algorithms for solving nonlinearly constrained optimization problems. Although second derivative information may often be calculated, there is little practical theory that justifies exact-Hessian SQP methods. In particular, the resulting quadratic programming (QP) subproblems are often nonconvex, and thus finding their global solutions may be computationally nonviable. This paper presents a second-derivative Sl1QP method based on quadratic subproblems that are either convex, and thus may be solved efficiently, or need not be solved globally. Additionally, an explicit descent constraint is imposed on certain QP subproblems, which “guides” the iterates through areas in which nonconvexity is a concern. Global convergence of the resulting algorithm is established.
منابع مشابه
A Second Derivative Sqp Method : Theoretical Issues
Sequential quadratic programming (SQP) methods form a class of highly efficient algorithms for solving nonlinearly constrained optimization problems. Although second derivative information may often be calculated, there is little practical theory that justifies exact-Hessian SQP methods. In particular, the resulting quadratic programming (QP) subproblems are often nonconvex, and thus finding th...
متن کاملUsing Modified IPSO-SQP Algorithm to Solve Nonlinear Time Optimal Bang-Bang Control Problem
In this paper, an intelligent-gradient based algorithm is proposed to solve time optimal bang-bang control problem. The proposed algorithm is a combination of an intelligent algorithm called improved particle swarm optimization algorithm (IPSO) in the first stage of optimization process together with a gradient-based algorithm called successive quadratic programming method (SQP) in the second s...
متن کاملA Globally Convergent Stabilized Sqp Method: Superlinear Convergence
Regularized and stabilized sequential quadratic programming (SQP) methods are two classes of methods designed to resolve the numerical and theoretical difficulties associated with ill-posed or degenerate nonlinear optimization problems. Recently, a regularized SQP method has been proposed that allows convergence to points satisfying certain second-order KKT conditions (SIAM J. Optim., 23(4):198...
متن کاملA Second Derivative Sqp Method : Local Convergence
In [19], we gave global convergence results for a second-derivative SQP method for minimizing the exact l1-merit function for a fixed value of the penalty parameter. To establish this result, we used the properties of the so-called Cauchy step, which was itself computed from the so-called predictor step. In addition, we allowed for the computation of a variety of (optional) SQP steps that were ...
متن کاملA Second Derivative SQP Method: Global Convergence
Gould and Robinson (NAR 08/18, Oxford University Computing Laboratory, 2008) gave global convergence results for a second-derivative SQP method for minimizing the exact l1-merit function for a fixed value of the penalty parameter. To establish this result, we used the properties of the so-called Cauchy step, which was itself computed from the so-called predictor step. In addition, we allowed fo...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2008